Distillation Sparsity Training Algorithm for Accelerating Convolutional Neural Networks in Embedded Systems

نویسندگان

چکیده

The rapid development of neural networks has come at the cost increased computational complexity. Neural are both computationally intensive and memory intensive; as such, minimal energy computing power satellites pose a challenge for automatic target recognition (ATR). Knowledge distillation (KD) can distill knowledge from cumbersome teacher network to lightweight student network, transferring essential information learned by network. Thus, concept KD be used improve accuracy networks. Even when learning there is still redundancy in Traditional fix structure before training, such that training does not situation. This paper proposes sparsity (DST) algorithm based on pruning address above limitations. We first through KD, then pruning, allowing learn which connections essential. DST allows teach pruned directly. proposed was tested CIFAR-100, MSTAR, FUSAR-Ship data sets, with 50% setting. First, new loss function teacher-pruned proposed, showed performance close Second, model (uniformity half-pruning UHP) designed solve problem unstructured facilitate implementation general-purpose hardware acceleration storage. Compared traditional UHP double speed

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Accelerating Convolutional Neural Network Systems

Convolutional Neural Networks have recently been shown to be highly effective classifiers for image and speech data. Due to the large volume of data required to build useful models, and the complexity of the models themselves, efficiency has become one of the primary concerns. This work shows that frequency domain methods can be utilised to accelerate the performance training, inference, and sl...

متن کامل

Towards dropout training for convolutional neural networks

Recently, dropout has seen increasing use in deep learning. For deep convolutional neural networks, dropout is known to work well in fully-connected layers. However, its effect in convolutional and pooling layers is still not clear. This paper demonstrates that max-pooling dropout is equivalent to randomly picking activation based on a multinomial distribution at training time. In light of this...

متن کامل

The Power of Sparsity in Convolutional Neural Networks

Deep convolutional networks are well-known for their high computational and memory demands. Given limited resources, how does one design a network that balances its size, training time, and prediction accuracy? A surprisingly effective approach to trade accuracy for size and speed is to simply reduce the number of channels in each convolutional layer by a fixed fraction and retrain the network....

متن کامل

Semi-Supervised Training of Convolutional Neural Networks

In this paper we discuss a method for semi-supervised training of CNNs. By using auto-encoders to extract features from unlabeled images, we can train CNNs to accurately classify images with only a small set of labeled images. We show our method’s results on a shallow CNN using the CIFAR-10 dataset, and some preliminary results on a VGG-16 network using the STL-10 dataset.

متن کامل

Implementation of Training Convolutional Neural Networks

Deep learning refers to a shining branch of machine learning that is based on learning levels of representations. Convolutional Neural Networks (CNN) is one kind of deep neural network. It can study concurrently. In this article, we use convolutional neural network to implement the typical face recognition problem which can overcome the influence of pose or resolution in face recognition. Then,...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Remote Sensing

سال: 2023

ISSN: ['2315-4632', '2315-4675']

DOI: https://doi.org/10.3390/rs15102609